On unlearnable problems -or- A model for premature saturation in backpropagation learning
نویسندگان
چکیده
In this paper we investigate the learning of an unlearnable problem and how this relates to the premature saturation of hidden neu-rons in error backpropagation learning. General aspects of our model are discussed. A sketch of the derivation of equations for the development of the signiicant weights in time is given. 1. Introduction The phenomenon of premature saturation of hidden neurons in feedforward neural networks trained by error backpropagation learning has repeatedly been reported by diierent researchers 2]. Diierent approaches have been proposed to circumvent this severe problem that can prevent proper learning. In 4] it is stated that the saturation is due to improperly chosen initial weights, where improper is to be regarded with respect to network parameters. We show that the relationship between these network parameters and the data to be learned is the major eeect leading to the undesirable growth of some weights. Therefore we will suggest and discuss a model for an extremly diicult learning task, relate it to backpropagation learning and then sketch the derivations of equations for the weight development during saturation. 2. The Model During many experiments reported elsewhere 1], we could observe that the probability for premature saturation depends on the relationship between the network parameters and the data, with which the network is to be trained. Especially when a problem is diicult to learn for a network (which does not imply that the chosen connguration is not well suited to accomplish the task), saturation can be observed. An extreme task that can never be learned by any This work has partially been funded by the German Federal Ministry education science research and technology as part of the AENEAS project, grand number 01 IN 505 C 4.
منابع مشابه
Unimodal Loading Problems
This paper deals with optimal learning and provides a uniied viewpoint of most signiicant results in the eld. The focus is on the problem of local minima in the cost function that is likely to aaect more or less any learning algorithm. We give some intriguing links between optimal learning and the computational complexity of loading problems. We exhibit a computational model such that the solut...
متن کاملAn analysis of premature saturation in back propagation learning
-The back propagation (BP) algorithm is widely used for finding optimum weights of multilayer neural networks in many pattern recognition applications. However, the critical drawbacks of the algorithm are its slow learning speed and convergence to local minima. One of the major reasons for these drawbacks is the "premature saturation'" which is a plwnomenon that the error of the neural network ...
متن کاملUnimodal Loading
This paper deals with optimal learning and provides a uniied viewpoint of most signiicant results in the eld. The focus is on the problem of local minima in the cost function that is likely to aaect more or less any learning algorithm. We give some intriguing links between optimal learning and the computational complexity of loading problems. We exhibit a computational model such that the solut...
متن کاملLoading Discriminative Feature Representations in Hidden Layer
This work explores the neural features that are trained by decreasing a discriminative energy. It directly resolves the unfaithful representation problem and the ambiguous internal representation problem in various backpropagation training algorithms for MLP. It also indirectly overcomes the premature saturation problem. Keywords—Multilayer perceptron; deep learning; Boltzmann machine; ambiguou...
متن کاملFast Back-Propagation Learning Using Steep Activation Functions and Automatic Weight Reinitialization
-In this paper, several back-propagation (BP) learning speed-up algorithms that employ the “gain” parameter, i.e., steepness of the activation function, are examined. Simulations will show that increasing the gain seemingly increases the speed of convergence and that these algorithms can converge faster than the standard BP learning algorithm on some problems. However, these algorithms may also...
متن کامل